Convergence Rate Analysis of the Forward-Douglas-Rachford Splitting Scheme
نویسنده
چکیده
Operator splitting schemes are a class of powerful algorithms that solve complicated monotone inclusion and convex optimization problems that are built from many simpler pieces. They give rise to algorithms in which all simple pieces of the decomposition are processed individually. This leads to easily implementable and highly parallelizable or distributed algorithms, which often obtain nearly state-of-the-art performance. In this paper, we analyze the convergence rate of the forward-Douglas-Rachford splitting (FDRS) algorithm, which is a generalization of the forward-backward splitting (FBS) and Douglas-Rachford splitting (DRS) algorithms. Under general convexity assumptions, we derive the ergodic and nonergodic convergence rates of the FDRS algorithm, and show that these rates are the best possible. Under Lipschitz differentiability assumptions, we show that the best iterate of FDRS converges as quickly as the last iterate of the FBS algorithm. Under strong convexity assumptions, we derive convergence rates for a sequence that strongly converges to a minimizer. Under strong convexity and Lipschitz differentiability assumptions, we show that FDRS converges linearly. We also provide examples where the objective is strongly convex, yet FDRS converges arbitrarily slowly. Finally, we relate the FDRS algorithm to a primal-dual forward-backward splitting scheme and clarify its place among existing splitting methods. Our results show that the FDRS algorithm automatically adapts to the regularity of the objective functions and achieves rates that improve upon the sharp worst case rates that hold in the absence of smoothness and strong convexity.
منابع مشابه
Stochastic Forward Douglas-Rachford Splitting for Monotone Inclusions
We propose a stochastic Forward Douglas-Rachford Splitting framework for finding a zero point of the sum of three maximally monotone operators in real separable Hilbert space, where one of them is cocoercive. We first prove the weak almost sure convergence of the proposed method. We then characterize the rate of convergence in expectation in the case of strongly monotone operators. Finally, we ...
متن کاملLecture 20 : Splitting Algorithms
In this lecture, we discuss splitting algorithms for convex minimization problems with objective given by the sum of two nonsmooth functions. We start with the fixed point property of such problems and derive a general scheme of splitting algorithm based on fixed point iteration. This covers Douglas-Rachford splitting and Peaceman-Rachford splitting algorithms. We also discuss the convergence r...
متن کاملA Douglas-Rachford splitting for semi-decentralized generalized Nash equilibrium seeking in Monotone Aggregative Games
We address the generalized Nash equilibrium seeking problem for noncooperative agents playing non-strictly monotone aggregative games with affine coupling constraints. First, we use operator theory to characterize the generalized Nash equilibria of the game as the zeros of a monotone setvalued operator. Then, we massage the Douglas–Rachford splitting to solve the monotone inclusion problem and ...
متن کاملMetric Selection in Douglas-Rachford Splitting and ADMM
Recently, several convergence rate results for Douglas-Rachford splitting and the alternating direction method of multipliers (ADMM) have been presented in the literature. In this paper, we show linear convergence of Douglas-Rachford splitting and ADMM under certain assumptions. We also show that the provided bounds on the linear convergence rates generalize and/or improve on similar bounds in ...
متن کاملOn convergence rate of the Douglas-Rachford operator splitting method
This note provides a simple proof on a O(1/k) convergence rate for the DouglasRachford operator splitting method where k denotes the iteration counter.
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- SIAM Journal on Optimization
دوره 25 شماره
صفحات -
تاریخ انتشار 2015